21 research outputs found
A Second-Order Boundary Value Problem with Nonlinear and Mixed Boundary Conditions: Existence, Uniqueness, and Approximation
A second-order boundary value problem with nonlinear and mixed two-point boundary conditions is considered, Lx=f(t,x,x′), t∈(a,b), g(x(a),x(b),x′(a),x′(b))=0, x(b)=x(a) in which L is a formally self-adjoint second-order differential operator. Under appropriate assumptions on L, f, and g, existence and uniqueness of solutions is established by the method of upper and lower solutions and Leray-Schauder degree theory. The general quasilinearization method is then applied to this problem. Two monotone sequences converging quadratically to the unique solution are constructed
Stable Learning via Sample Reweighting
We consider the problem of learning linear prediction models with model
misspecification bias. In such case, the collinearity among input variables may
inflate the error of parameter estimation, resulting in instability of
prediction results when training and test distributions do not match. In this
paper we theoretically analyze this fundamental problem and propose a sample
reweighting method that reduces collinearity among input variables. Our method
can be seen as a pretreatment of data to improve the condition of design
matrix, and it can then be combined with any standard learning method for
parameter estimation and variable selection. Empirical studies on both
simulation and real datasets demonstrate the effectiveness of our method in
terms of more stable performance across different distributed data.Comment: Accepted as poster paper at AAAI202
Distributionally Robust Learning with Stable Adversarial Training
Machine learning algorithms with empirical risk minimization are vulnerable
under distributional shifts due to the greedy adoption of all the correlations
found in training data. There is an emerging literature on tackling this
problem by minimizing the worst-case risk over an uncertainty set. However,
existing methods mostly construct ambiguity sets by treating all variables
equally regardless of the stability of their correlations with the target,
resulting in the overwhelmingly-large uncertainty set and low confidence of the
learner. In this paper, we propose a novel Stable Adversarial Learning (SAL)
algorithm that leverages heterogeneous data sources to construct a more
practical uncertainty set and conduct differentiated robustness optimization,
where covariates are differentiated according to the stability of their
correlations with the target. We theoretically show that our method is
tractable for stochastic gradient-based optimization and provide the
performance guarantees for our method. Empirical studies on both simulation and
real datasets validate the effectiveness of our method in terms of uniformly
good performance across unknown distributional shifts.Comment: arXiv admin note: substantial text overlap with arXiv:2006.0441
Relaxation Oscillations in Singularly Perturbed Generalized Lienard Systems with Non-Generic Turning Points
Based on the asymptotic analysis technique developed by Eckhaus [Lecture Notes in Math., vol. 985, pp 449-494. Springer, Berlin, 1983], this paper aims to study the existence and the asymptotic behaviors of relaxation oscillations of regular and canard types in a singularly perturbed generalized Lionard system with a non-generic turning point. The singularly perturbed Lionard system considered in this paper is very general and numerous real world models like some biological ones can be rewritten in the form of this system after a series of transformations. Under certain conditions, we rigorously prove the existence of regular relaxation oscillations and canard relaxation oscillations under the specific parameter conditions. As an application, two biological models, namely, a FitzHugh-Nagumo model and a twodimensional predator-prey model with Holling-II response are studied, in which, the existence of regular relaxation oscillations and canard relaxation oscillations as well as the bifurcation curves are obtained
Stable Learning via Sample Reweighting
We consider the problem of learning linear prediction models with model misspecification bias. In such case, the collinearity among input variables may inflate the error of parameter estimation, resulting in instability of prediction results when training and test distributions do not match. In this paper we theoretically analyze this fundamental problem and propose a sample reweighting method that reduces collinearity among input variables. Our method can be seen as a pretreatment of data to improve the condition of design matrix, and it can then be combined with any standard learning method for parameter estimation and variable selection. Empirical studies on both simulation and real datasets demonstrate the effectiveness of our method in terms of more stable performance across different distributed data